The Thing We Tried That Worked: Utile Distinctions for Relational Reinforcement Learning

نویسندگان

  • William Dabney
  • Amy McGovern
چکیده

This paper introduces a relational function approximation technique based on McCallum’s UTree algorithm (McCallum, 1995). We have extended the original approach to handle relational observations using an attribute graph of observable objects and relationships (McGovern et al., 2003). Furthermore, we address the inherent challenges that arise with a relational representation. We use stochastic sampling to manage the search space (Srinivasan, 1999), and sampling to address issues of autocorrelation (Jensen and Neville, 2002). We prevent the algorithm from growing an overly large and complex tree by incorporating Iterative Tree Induction’s approach (Utgoff, 1995). We compare Relational UTree’s performance with similar relational learning methods (Finney et al., 2002) (Driessens et al., 2001).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Utile Distinctions for Relational Reinforcement Learning

We introduce an approach to autonomously creating state space abstractions for an online reinforcement learning agent using a relational representation. Our approach uses a tree-based function approximation derived from McCallum’s [1995] UTree algorithm. We have extended this approach to use a relational representation where relational observations are represented by attributed graphs [McGovern...

متن کامل

Instance-Based Utile Distinctions for Reinforcement Learning with Hidden State

We present Utile Suffix Memory, a reinforcement learning algorithm that uses short-term memory to overcome the state aliasing that results from hidden state. By combining the advantages of previous work in instance-based (or “memorybased”) learning and previous work with statistical tests for separating noise from task structure, the method learns quickly, creates only as much memory as needed ...

متن کامل

MULTI-MODAL UTILE DISTINCTIONS Multi-Modal Utile Distinctions

We introduce Multi-Modal Utility Trees (MMU), an algorithm for autonomously learning decision treebased state abstractions in Partially Observable Markov Decision Processes with multi-modal observations. MMU builds the trees using the Kolmogorov-Smirnov statistical test. Additionally, MMU incorporates the ability to perform online tree restructuring, enabling it to build and maintain a compact ...

متن کامل

Learning to Use Selective Attention and Short-Term Memory in Sequential Tasks

This paper presents U-Tree, a reinforcement learning algorithm that uses selective attention and shortterm memory to simultaneously address the intertwined problems of large perceptual state spaces and hidden state. By combining the advantages of work in instance-based (or “memory-based”) learning and work with robust statistical tests for separating noise from task structure, the method learns...

متن کامل

Efficient Exploration in Reinforcement Learning Based on Utile Suffix Memory

Reinforcement learning addresses the question of how an autonomous agent can learn to choose optimal actions to achieve its goals. Efficient exploration is of fundamental importance for autonomous agents that learn to act. Previous approaches to exploration in reinforcement learning usually address exploration in the case when the environment is fully observable. In contrast, we study the case ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006